underwater environment
Autonomous Underwater Cognitive System for Adaptive Navigation: A SLAM-Integrated Cognitive Architecture
Jayarathne, K. A. I. N, Rathnayaka, R. M. N. M., Peiris, D. P. S. S.
Abstract--Deep-sea exploration faces critical challenges including disorientation, communication loss, and navigational failures in hostile underwater environments. This paper presents an Autonomous Underwater Cognitive System (AUCS) that integrates Simultaneous Localization and Mapping (SLAM) with a Soar-based cognitive architecture to enable adaptive navigation under dynamic oceanic conditions. The system combines multi-sensor fusion (SONAR, LiDAR, IMU, DVL) with cognitive reasoning capabilities including perception, attention, planning, and learning. Unlike conventional reactive SLAM systems, AUCS incorporates semantic understanding, adaptive sensor management, and memory-based learning to distinguish between dynamic and static objects, thus reducing false loop closures and improving long-term map consistency. This work addresses critical safety limitations observed in previous deep-sea missions and establishes a foundation for next-generation cognitive submersible systems.
When Semantics Connect the Swarm: LLM-Driven Fuzzy Control for Cooperative Multi-Robot Underwater Coverage
Xu, Jingzehua, Zhang, Weihang, Li, Yangyang, Zhang, Hongmiaoyi, Xie, Guanwen, Tang, Jiwei, Zhang, Shuai, Li, Yi
Underwater multi-robot cooperative coverage remains challenging due to partial observability, limited communication, environmental uncertainty, and the lack of access to global localization. To address these issues, this paper presents a semantics-guided fuzzy control framework that couples Large Language Models (LLMs) with interpretable control and lightweight coordination. Raw multimodal observations are compressed by the LLM into compact, human-interpretable semantic tokens that summarize obstacles, unexplored regions, and Objects Of Interest (OOIs) under uncertain perception. A fuzzy inference system with pre-defined membership functions then maps these tokens into smooth and stable steering and gait commands, enabling reliable navigation without relying on global positioning. Then, we further coordinate multiple robots by introducing semantic communication that shares intent and local context in linguistic form, enabling agreement on who explores where while avoiding redundant revisits. Extensive simulations in unknown reef-like environments show that, under limited sensing and communication, the proposed framework achieves robust OOI-oriented navigation and cooperative coverage with improved efficiency and adaptability, narrowing the gap between semantic cognition and distributed underwater control in GPS-denied, map-free conditions.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- Asia > China > Hong Kong (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- (4 more...)
- Overview (0.68)
- Research Report (0.50)
SonarSweep: Fusing Sonar and Vision for Robust 3D Reconstruction via Plane Sweeping
Chen, Lingpeng, Tang, Jiakun, Chui, Apple Pui-Yi, Hong, Ziyang, Wu, Junfeng
Accurate 3D reconstruction in visually-degraded underwater environments remains a formidable challenge. Single-modality approaches are insufficient: vision-based methods fail due to poor visibility and geometric constraints, while sonar is crippled by inherent elevation ambiguity and low resolution. Consequently, prior fusion technique relies on heuristics and flawed geometric assumptions, leading to significant artifacts and an inability to model complex scenes. In this paper, we introduce SonarSweep, a novel, end-to-end deep learning framework that overcomes these limitations by adapting the principled plane sweep algorithm for cross-modal fusion between sonar and visual data. Extensive experiments in both high-fidelity simulation and real-world environments demonstrate that SonarSweep consistently generates dense and accurate depth maps, significantly outperforming state-of-the-art methods across challenging conditions, particularly in high turbidity. To foster further research, we will publicly release our code and a novel dataset featuring synchronized stereo-camera and sonar data, the first of its kind.
- North America > United States (0.04)
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
EREBUS: End-to-end Robust Event Based Underwater Simulation
Kyatham, Hitesh, Suresh, Arjun, Palnitkar, Aadi, Aloimonos, Yiannis
Abstract--The underwater domain presents a vast array of challenges for roboticists and computer vision researchers alike, such as poor lighting conditions and high dynamic range scenes. In these adverse conditions, traditional vision techniques struggle to adapt and lead to suboptimal performance. Event-based cameras present an attractive solution to this problem, mitigating the issues of traditional cameras by tracking changes in the footage on a frame-by-frame basis. In this paper, we introduce a pipeline which can be used to generate realistic synthetic data of an event-based camera mounted to an AUV (Autonomous Underwater V ehicle) in an underwater environment for training vision models. We demonstrate the effectiveness of our pipeline using the task of rock detection with poor visibility and suspended particulate matter, but the approach can be generalized to other underwater tasks.
- North America > United States > Maryland > Prince George's County > College Park (0.14)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
Enhancing Underwater Object Detection through Spatio-Temporal Analysis and Spatial Attention Networks
Karri, Sai Likhith, Saxena, Ansh
This study examines the effectiveness of spatio-temporal modeling and the integration of spatial attention mechanisms in deep learning models for underwater object detection. Specifically, in the first phase, the performance of temporal-enhanced YOLOv5 variant T-YOLOv5 is evaluated, in comparison with the standard YOLOv5. For the second phase, an augmented version of T-YOLOv5 is developed, through the addition of a Convolutional Block Attention Module (CBAM). By examining the effectiveness of the already pre-existing YOLOv5 and T-YOLOv5 models and of the newly developed T-YOLOv5 with CBAM. With CBAM, the research highlights how temporal modeling improves detection accuracy in dynamic marine environments, particularly under conditions of sudden movements, partial occlusions, and gradual motion. The testing results showed that YOLOv5 achieved a mAP@50-95 of 0.563, while T-YOLOv5 and T-YOLOv5 with CBAM outperformed with mAP@50-95 scores of 0.813 and 0.811, respectively, highlighting their superior accuracy and generalization in detecting complex objects. The findings demonstrate that T-YOLOv5 significantly enhances detection reliability compared to the standard model, while T-YOLOv5 with CBAM further improves performance in challenging scenarios, although there is a loss of accuracy when it comes to simpler scenarios.
Advancing Marine Research: UWSAM Framework and UIIS10K Dataset for Precise Underwater Instance Segmentation
Li, Hua, Lian, Shijie, Li, Zhiyuan, Cong, Runmin, Li, Chongyi, Yang, Laurence T., Zhang, Weidong, Kwong, Sam
With recent breakthroughs in large-scale modeling, the Segment Anything Model (SAM) has demonstrated significant potential in a variety of visual applications. However, due to the lack of underwater domain expertise, SAM and its variants face performance limitations in end-to-end underwater instance segmentation tasks, while their higher computational requirements further hinder their application in underwater scenarios. To address this challenge, we propose a large-scale underwater instance segmentation dataset, UIIS10K, which includes 10,048 images with pixel-level annotations for 10 categories. Then, we introduce UWSAM, an efficient model designed for automatic and accurate segmentation of underwater instances. UWSAM efficiently distills knowledge from the SAM ViT-Huge image encoder into the smaller ViT-Small image encoder via the Mask GAT-based Underwater Knowledge Distillation (MG-UKD) method for effective visual representation learning. Furthermore, we design an End-to-end Underwater Prompt Generator (EUPG) for UWSAM, which automatically generates underwater prompts instead of explicitly providing foreground points or boxes as prompts, thus enabling the network to locate underwater instances accurately for efficient segmentation. Comprehensive experimental results show that our model is effective, achieving significant performance improvements over state-of-the-art methods on multiple underwater instance datasets. Datasets and codes are available at https://github.com/LiamLian0727/UIIS10K.
- Europe > Switzerland (0.04)
- Asia > China > Shanghai > Shanghai (0.04)
- Asia > China > Hubei Province > Wuhan (0.04)
- (7 more...)
- Research Report > New Finding (0.34)
- Research Report > Promising Solution (0.34)
A Structured Review of Underwater Object Detection Challenges and Solutions: From Traditional to Large Vision Language Models
Nabahirwa, Edwine, Song, Wei, Zhang, Minghua, Fang, Yi, Ni, Zhou
Despite its significance, the underwater world remains largely overlooked as a result of the challenging conditions that hinder traditional research methods. Historically, the study of marine ecosystems relied on labor intensive research [1], which provided limited data and had a high error margin. In recent years, advances in autonomous and remotely operated vehicles (AUVs and ROVs) have revolutionized underwater exploration. These technologies, equipped with object detection systems, now allow real-time monitoring, which includes capturing images of marine organisms, environmental conditions, and even assessing biodiversity [2], [3]. However, the quality of images and videos captured underwater remains a significant obstacle. Light absorption, scattering, and water-related distortions, such as haze and color shifts [4], create noisy low-contrast images, further compounded by complex underwater backgrounds and camera motion. These challenges call for advanced detection techniques capable of accurately identifying and localizing objects despite underwater noise. Efficient underwater object detection (UOD) is crucial for a variety of marine applications, including biodiversity monitoring, conservation efforts, and resource management.
- Asia > China > Shanghai > Shanghai (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- Asia > Japan > Honshū > Kantō > Ibaraki Prefecture > Tsukuba (0.04)
- (20 more...)
- Overview (1.00)
- Research Report > Promising Solution (0.92)
- Media > Photography (0.48)
- Health & Medicine > Diagnostic Medicine (0.45)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Vision > Image Understanding (0.92)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.47)
DRACo-SLAM2: Distributed Robust Acoustic Communication-efficient SLAM for Imaging Sonar EquippedUnderwater Robot Teams with Object Graph Matching
Huang, Yewei, McConnell, John, Lin, Xi, Englot, Brendan
We present DRACo-SLAM2, a distributed SLAM framework for underwater robot teams equipped with multibeam imaging sonar. This framework improves upon the original DRACo-SLAM by introducing a novel representation of sonar maps as object graphs and utilizing object graph matching to achieve time-efficient inter-robot loop closure detection without relying on prior geometric information. To better-accommodate the needs and characteristics of underwater scan matching, we propose incremental Group-wise Consistent Measurement Set Maximization (GCM), a modification of Pairwise Consistent Measurement Set Maximization (PCM), which effectively handles scenarios where nearby inter-robot loop closures share similar registration errors. The proposed approach is validated through extensive comparative analyses on simulated and real-world datasets.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > New Jersey > Hudson County > Hoboken (0.04)
- North America > United States > Maryland > Anne Arundel County > Annapolis (0.04)
- Energy (0.68)
- Leisure & Entertainment > Games (0.61)
- Transportation (0.47)
Context-Aware Behavior Learning with Heuristic Motion Memory for Underwater Manipulation
Buchholz, Markus, Carlucho, Ignacio, Grimaldi, Michele, Koskinopoulou, Maria, Petillot, Yvan R.
Autonomous motion planning is critical for efficient and safe underwater manipulation in dynamic marine environments. Current motion planning methods often fail to effectively utilize prior motion experiences and adapt to real-time uncertainties inherent in underwater settings. In this paper, we introduce an Adaptive Heuristic Motion Planner framework that integrates a Heuristic Motion Space (HMS) with Bayesian Networks to enhance motion planning for autonomous underwater manipulation. Our approach employs the Probabilistic Roadmap (PRM) algorithm within HMS to optimize paths by minimizing a composite cost function that accounts for distance, uncertainty, energy consumption, and execution time. By leveraging HMS, our framework significantly reduces the search space, thereby boosting computational performance and enabling real-time planning capabilities. Bayesian Networks are utilized to dynamically update uncertainty estimates based on real-time sensor data and environmental conditions, thereby refining the joint probability of path success. Through extensive simulations and real-world test scenarios, we showcase the advantages of our method in terms of enhanced performance and robustness. This probabilistic approach significantly advances the capability of autonomous underwater robots, ensuring optimized motion planning in the face of dynamic marine challenges.
- North America > United States (0.04)
- Europe > Switzerland (0.04)
- Asia > Singapore (0.04)
- Information Technology > Artificial Intelligence > Robots > Robot Planning & Action (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.90)
Stable Acoustic Relay Assignment with High Throughput via Lase Chaos-based Reinforcement Learning
Chen, Zengjing, Wang, Lu, Xing, Chengzhi
Underwater Acoustic Networks (UANs) have gained significant attention from both industry and academia due to their indisputable advantages in improving link reliability, increasing system capacity, expanding transmission range and so on. Acoustic communication is most widely used underwater communication as sound wave is not absorbed by water so easily like electromagnetic wave and optical wave [1]. UANs typically consist of acoustic-linked seabed sensors, autonomous underwater vehicles, and ground stations that provide links to onshore control centers. Due to the battery-powered network nodes, shallow water acoustic channel characteristics, such as low available bandwidth and highly varying multi-path, maximizing throughput while minimizing consumption has become a very challenging task [2]. Recent studies have discussed the challenges and opportunities of underwater cognitive communication [3], proposed cooperative automatic repeat request protocols for higher channel quality [4], and analyzed the impact of low transmission rates and long preambles on medium access control protocols [5]. Artificial intelligence (AI) has experienced significant growth in popularity in recent years, and many industries and research fields have explored its potential applications, including information theory, game theory, biological systems, and so on [6-9].
- Energy (0.48)
- Information Technology > Security & Privacy (0.34)
- Information Technology > Communications > Networks (1.00)
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)